A Triangle Inequality for Covariances of Binary FKG Random Variables

نویسندگان
چکیده

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fano's inequality for random variables

We extend Fano’s inequality, which controls the average probability of (disjoint) events in terms of the average of some Kullback-Leibler divergences, to work with arbitrary [0, 1]–valued random variables. Our simple two-step methodology is general enough to cover the case of an arbitrary (possibly continuously infinite) family of distributions as well as [0, 1]–valued random variables not nece...

متن کامل

Bounds for covariances and variances of truncated random variables

We show that a lower bound for covariance of min(X 1 , X 2) and max(X 1 , X 2) is cov (X 1 , X 2) and an upper bound for variance of min(X 2 , max(X, X 1)) is var (X) + var (X 1) + var (X 2) generalizing previous results. We also characterize the cases where these bounds are sharp.

متن کامل

The FKG Inequality for Partially Ordered Algebras

The FKG inequality asserts that for a distributive lattice with logsupermodular probability measure, any two increasing functions are positively correlated. In this paper we extend this result to functions with values in partially ordered algebras, such as algebras of matrices and polynomials.

متن کامل

An entropy inequality for symmetric random variables

We establish a lower bound on the entropy of weighted sums of (possibly dependent) random variables (X1, X2, . . . , Xn) possessing a symmetric joint distribution. Our lower bound is in terms of the joint entropy of (X1, X2, . . . , Xn). We show that for n ≥ 3, the lower bound is tight if and only if Xi’s are i.i.d. Gaussian random variables. For n = 2 there are numerous other cases of equality...

متن کامل

Symmetrization of Binary Random Variables

A random variable Y is called an independent symmetrizer of a given random variable X if (a) it is independent of X and (b) the distribution of X + Y is symmetric about 0. In cases where the distribution of X is symmetric about its mean, it is easy to see that the constant random variable Y = −EX is a minimum-variance independent symmetrizer. Taking Y to have the same distribution as −X clearly...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: The Annals of Applied Probability

سال: 1995

ISSN: 1050-5164

DOI: 10.1214/aoap/1177004843